(Stochastic gradient descent (SGD))
Efficient offline optimization method for functions ff with finite sum structure: f(𝐱)=∑i=1nfi(𝐱)f(\mathbf{x})=\sum_{i=1}^n f_i(\mathbf{x}) Goal is to find 𝐱̂\hat{\mathbf{x}} such that
Gradient descent
#incomplete
References: 1. 2. https://web.stanford.edu/class/ee270/scribes/lecture16.pdf